Experts Combination through Density Decomposition
نویسندگان
چکیده
This paper is concerned with an important issue in Statistics and Artiicial Intelligence, which is problem decomposition and experts (or predictors) combination. Decomposition methods usually adopt a divide-and-conquer strategy which decomposes the initial problem into simple sub-problems. The global expert is then obtained from some combination of the local experts. In the case of Hard decomposition, the local experts partition the input space without overlap, while in soft decomposition, experts overlap and may specialize on a common region of the input space. The overlap usually makes combination more powerful, since it introduces redundancy and smoothness in the experts outputs. On the other hand, there is a considerable literature on experts combination (independently of the decomposition) which consists in combining several global experts. Most of the studies deal with linear combination whose theory is better known and more tractable than the nonlin-ear case. There are several methods for linear combination which have been shown to improve performance by signiicantly reducing variancee3]]5]]1]. Although theoretical requirements for linear combination are not usually fulllled, because of the sample size and the statistical dependencies between individual experts, experimental results usually show improved performance over a single expert even when being chosen according to a cross-validation process. Actually, linear combination can be seen as a cross-validation in the functional (experts) space. The nal result is an average expert, while classical cross-validation (in parameters space) usually leads to some average performance and leaves the user choose a single expert according to some strategy (e.g. choose expert with the best performance). In this paper, we adopt a divide-and-combine strategy to solve a complex prediction task. The methodology consists in the following steps: 1. Assume the input density is a mixture of prior densities and estimate its parameters from the data. For our study we consider Gaussian distributions. 2. Decompose the global problem into a number of sub-problems according to a soft division process which is driven by the input density. The number of sub-problems is the number of Gaussians (clusters) in the mixture. 3. To every Gaussian (cluster), assign a local expert which is trained on the corresponding data (for which density is greater than a threshold, a small threshold allows more overlap between Gaussians of the mixture). 4. Combine the local experts. Determination of parameters of the combination is performed with two methods: i) Stacking, which consists in estimating weights which minimize some 1
منابع مشابه
Local experts combination through density decomposition
In this paper we describe a divide-and-combine strategy for decomposition of a complex prediction problem into simpler local sub-problems. We rstly show how to perform a soft decomposition via clustering of input data. Such decomposition leads to a partition of the input space into several regions which may overlap. Therefore, to each region is assigned a local predictor (or expert) which is tr...
متن کاملSolving Differential Equations by Using a Combination of the First Kind Chebyshev Polynomials and Adomian Decomposition Method
In this paper, we are going to solve a class of ordinary differential equations that its source term are rational functions. We obtain the best approximation of source term by Chebyshev polynomials of the first kind, then we solve the ordinary differential equations by using the Adomian decomposition method
متن کاملProgressive FastICA Peel-Off and Convolution Kernel Compensation Demonstrate High Agreement for High Density Surface EMG Decomposition
Decomposition of electromyograms (EMG) is a key approach to investigating motor unit plasticity. Various signal processing techniques have been developed for high density surface EMG decomposition, among which the convolution kernel compensation (CKC) has achieved high decomposition yield with extensive validation. Very recently, a progressive FastICA peel-off (PFP) framework has also been deve...
متن کاملMultiscale Dictionary Learning for Estimating Conditional Distributions
Nonparametric estimation of the conditional distribution of a response given highdimensional features is a challenging problem. It is important to allow not only the mean but also the variance and shape of the response density to change flexibly with features, which are massive-dimensional. We propose a multiscale dictionary learning model, which expresses the conditional response density as a ...
متن کاملGeneralized Product of Experts for Automatic and Principled Fusion of Gaussian Process Predictions
In this work, we propose a generalized product of experts (gPoE) framework for combining the predictions of multiple probabilistic models. We identify four desirable properties that are important for scalability, expressiveness and robustness, when learning and inferring with a combination of multiple models. Through analysis and experiments, we show that gPoE of Gaussian processes (GP) have th...
متن کامل